Goto

Collaborating Authors

 human body


Explore the human body in stunning, 3D detail with a new online tool

Popular Science

The free Human Organ Atlas gives users an up-close-and-personal look at 56 human organs. The Human Organ Atlas portal is open-access and includes the kidneys, brain, heart, and more. Breakthroughs, discoveries, and DIY tips sent six days a week. If watching is giving you a renewed interest in the human body in all of its gory glory, there's a new tool that will help satisfy your curiosity. An international team of scientists developed an open-access 3D portal where users can explore human organs in detail.


Sounding Bodies: Modeling 3D Spatial Sound of Humans Using Body Pose and Audio

Neural Information Processing Systems

The system consumes, as input, audio signals from headset microphones and body pose, and produces, as output, a 3D sound field surrounding the transmitter's






MIT Technology Review's most popular stories of 2025

MIT Technology Review

This year, hype around AI really exploded, and so did concerns about AI's environmental footprint. We also saw some surprising biotech developments. It's been a busy and productive year here at . We published magazine issues on power, creativity, innovation, bodies, relationships, and security . We hosted 14 exclusive virtual conversations with our editors and outside experts in our subscriber-only series, Roundtables, and held two events on MIT's campus. And we published hundreds of articles online, following new developments in computing, climate tech, robotics, and more.


Sounding Bodies: Modeling 3D Spatial Sound of Humans Using Body Pose and Audio

Neural Information Processing Systems

While 3D human body modeling has received much attention in computer vision, modeling the acoustic equivalent, i.e. modeling 3D spatial audio produced by body motion and speech, has fallen short in the community. To close this gap, we present a model that can generate accurate 3D spatial audio for full human bodies. The system consumes, as input, audio signals from headset microphones and body pose, and produces, as output, a 3D sound field surrounding the transmitter's body, from which spatial audio can be rendered at any arbitrary position in the 3D space. We collect a first-of-its-kind multimodal dataset of human bodies, recorded with multiple cameras and a spherical array of 345 microphones. In an empirical evaluation, we demonstrate that our model can produce accurate body-induced sound fields when trained with a suitable loss. Dataset and code are available online.


Garment4D: Garment Reconstruction from Point Cloud Sequences

Neural Information Processing Systems

Learning to reconstruct 3D garments is important for dressing 3D human bodies of different shapes in different poses. Previous works typically rely on 2D images as input, which however suffer from the scale and pose ambiguities. To circumvent the problems caused by 2D images, we propose a principled framework, Garment4D, that uses 3D point cloud sequences of dressed humans for garment reconstruction. Garment4D has three dedicated steps: sequential garments registration, canonical garment estimation, and posed garment reconstruction. The main challenges are two-fold: 1) effective 3D feature learning for fine details, and 2) capture of garment dynamics caused by the interaction between garments and the human body, especially for loose garments like skirts. To unravel these problems, we introduce a novel Proposal-Guided Hierarchical Feature Network and Iterative Graph Convolution Network, which integrate both high-level semantic features and low-level geometric features for fine details reconstruction. Furthermore, we propose a Temporal Transformer for smooth garment motions capture. Unlike non-parametric methods, the reconstructed garment meshes by our method are separable from the human body and have strong interpretability, which is desirable for downstream tasks. As the first attempt at this task, high-quality reconstruction results are qualitatively and quantitatively illustrated through extensive experiments.


Dense Correspondences between Human Bodies via Learning Transformation Synchronization on Graphs

Neural Information Processing Systems

We introduce an approach for establishing dense correspondences between partial scans of human models and a complete template model. Our approach's key novelty lies in formulating dense correspondence computation as initializing and synchronizing local transformations between the scan and the template model. We introduce an optimization formulation for synchronizing transformations among a graph of the input scan, which automatically enforces smoothness of correspondences and recovers the underlying articulated deformations. We then show how to convert the iterative optimization procedure among a graph of the input scan into an end-to-end trainable network. The network design utilizes additional trainable parameters to break the barrier of the original optimization formulation's exact and robust recovery conditions. Experimental results on benchmark datasets demonstrate that our approach considerably outperforms baseline approaches in accuracy and robustness.